/art/compiler/utils/arm64/ |
H A D | assembler_arm64.cc | 57 void Arm64Assembler::GetCurrentThread(FrameOffset offset, ManagedRegister /* scratch */) { 158 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 159 CHECK(scratch.IsXRegister()) << scratch; 160 LoadImmediate(scratch.AsXRegister(), imm); 161 StoreWToOffset(kStoreWord, scratch.AsOverlappingWRegister(), SP, 167 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 168 CHECK(scratch.IsXRegister()) << scratch; 169 LoadImmediate(scratch 176 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 192 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 358 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 367 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 375 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 385 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 401 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 420 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 445 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 515 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 524 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 535 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 576 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 611 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local [all...] |
H A D | assembler_arm64.h | 104 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 105 void StoreImmediateToThread64(ThreadOffset<8> dest, uint32_t imm, ManagedRegister scratch) 108 ManagedRegister scratch) OVERRIDE; 111 ManagedRegister scratch) OVERRIDE; 125 ManagedRegister scratch) OVERRIDE; 126 void CopyRawPtrToThread64(ThreadOffset<8> thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 128 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 129 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 130 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 132 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, [all...] |
/art/compiler/utils/arm/ |
H A D | assembler_arm.cc | 518 ArmManagedRegister scratch = mscratch.AsArm(); 520 LoadFromOffset(kLoadWord, scratch.AsCoreRegister(), SP, in_off.Int32Value()); 521 StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value() + 4); 526 ArmManagedRegister scratch = mscratch.AsArm(); 527 LoadFromOffset(kLoadWord, scratch.AsCoreRegister(), SP, src.Int32Value()); 528 StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value()); 558 ArmManagedRegister scratch = mscratch.AsArm(); 559 CHECK(scratch.IsCoreRegister()) << scratch; 560 LoadImmediate(scratch [all...] |
H A D | assembler_arm.h | 678 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 680 void StoreImmediateToThread32(ThreadOffset<4> dest, uint32_t imm, ManagedRegister scratch) 684 ManagedRegister scratch) OVERRIDE; 689 ManagedRegister scratch) OVERRIDE; 709 ManagedRegister scratch) OVERRIDE; 711 void CopyRawPtrToThread32(ThreadOffset<4> thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 714 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 716 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 718 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 721 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, 792 ArmExceptionSlowPath(ArmManagedRegister scratch, size_t stack_adjust) argument [all...] |
H A D | assembler_arm32.h | 291 void MemoryBarrier(ManagedRegister scratch) OVERRIDE;
|
H A D | assembler_thumb2.h | 331 void MemoryBarrier(ManagedRegister scratch) OVERRIDE;
|
/art/compiler/utils/ |
H A D | swap_space_test.cc | 35 ScratchFile scratch; local 36 int fd = scratch.GetFd(); 37 unlink(scratch.GetFilename().c_str()); 70 scratch.Close();
|
H A D | assembler.h | 411 ManagedRegister scratch) = 0; 414 ManagedRegister scratch); 416 ManagedRegister scratch); 420 ManagedRegister scratch); 423 ManagedRegister scratch); 429 FrameOffset in_off, ManagedRegister scratch) = 0; 451 ManagedRegister scratch); 453 ManagedRegister scratch); 456 ManagedRegister scratch); 458 ManagedRegister scratch); [all...] |
H A D | assembler.cc | 136 ManagedRegister scratch ATTRIBUTE_UNUSED) { 142 ManagedRegister scratch ATTRIBUTE_UNUSED) { 148 ManagedRegister scratch ATTRIBUTE_UNUSED) { 154 ManagedRegister scratch ATTRIBUTE_UNUSED) { 190 ManagedRegister scratch ATTRIBUTE_UNUSED) { 196 ManagedRegister scratch ATTRIBUTE_UNUSED) { 202 ManagedRegister scratch ATTRIBUTE_UNUSED) { 208 ManagedRegister scratch ATTRIBUTE_UNUSED) { 213 ManagedRegister scratch ATTRIBUTE_UNUSED) { 218 ManagedRegister scratch ATTRIBUTE_UNUSE [all...] |
/art/compiler/utils/mips/ |
H A D | assembler_mips.cc | 648 MipsManagedRegister scratch = mscratch.AsMips(); local 649 CHECK(scratch.IsCoreRegister()) << scratch; 650 LoadImmediate(scratch.AsCoreRegister(), imm); 651 StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value()); 656 MipsManagedRegister scratch = mscratch.AsMips(); local 657 CHECK(scratch.IsCoreRegister()) << scratch; 658 LoadImmediate(scratch.AsCoreRegister(), imm); 659 StoreToOffset(kStoreWord, scratch 665 MipsManagedRegister scratch = mscratch.AsMips(); local 679 MipsManagedRegister scratch = mscratch.AsMips(); local 763 MipsManagedRegister scratch = mscratch.AsMips(); local 772 MipsManagedRegister scratch = mscratch.AsMips(); local 783 MipsManagedRegister scratch = mscratch.AsMips(); local 793 MipsManagedRegister scratch = mscratch.AsMips(); local 809 Register scratch = mscratch.AsMips().AsCoreRegister(); local 817 Register scratch = mscratch.AsMips().AsCoreRegister(); local 832 Register scratch = mscratch.AsMips().AsCoreRegister(); local 878 MipsManagedRegister scratch = mscratch.AsMips(); local 923 MipsManagedRegister scratch = mscratch.AsMips(); local 933 MipsManagedRegister scratch = mscratch.AsMips(); local 958 MipsManagedRegister scratch = mscratch.AsMips(); local [all...] |
H A D | assembler_mips.h | 286 explicit MipsExceptionSlowPath(MipsManagedRegister scratch, size_t stack_adjust) argument 287 : scratch_(scratch), stack_adjust_(stack_adjust) {}
|
/art/compiler/utils/mips64/ |
H A D | assembler_mips64.cc | 1192 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1193 CHECK(scratch.IsGpuRegister()) << scratch; 1194 LoadConst32(scratch.AsGpuRegister(), imm); 1195 StoreToOffset(kStoreWord, scratch.AsGpuRegister(), SP, dest.Int32Value()); 1200 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1201 CHECK(scratch.IsGpuRegister()) << scratch; 1204 LoadConst32(scratch.AsGpuRegister(), imm); 1205 StoreToOffset(kStoreDoubleword, scratch 1211 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1224 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1304 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1313 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1322 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1332 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1348 GpuRegister scratch = mscratch.AsMips64().AsGpuRegister(); local 1365 GpuRegister scratch = mscratch.AsMips64().AsGpuRegister(); local 1388 GpuRegister scratch = mscratch.AsMips64().AsGpuRegister(); local 1447 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1492 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1502 Mips64ManagedRegister scratch = mscratch.AsMips64(); local 1527 Mips64ManagedRegister scratch = mscratch.AsMips64(); local [all...] |
H A D | assembler_mips64.h | 357 explicit Mips64ExceptionSlowPath(Mips64ManagedRegister scratch, size_t stack_adjust) argument 358 : scratch_(scratch), stack_adjust_(stack_adjust) {}
|
/art/compiler/optimizing/ |
H A D | parallel_move_resolver.cc | 57 // Eliminate the move, in case following moves need a scratch register. 251 int scratch = -1; local 254 scratch = reg; 259 if (scratch == -1) { 261 scratch = if_scratch; 266 return scratch; 390 // in the move graph. We use scratch location to resolve cycles, also 402 // resolve as (1 -> scratch) (0 -> 1) (scratch -> 0). If, by chance, '2' is 403 // used as the scratch locatio 446 Location scratch = AllocateScratchLocationFor(kind); local 480 Location scratch = scratches_.Get(i); local [all...] |
H A D | parallel_move_test.cc | 115 Location scratch = GetScratchLocation(kind); variable 116 if (scratch.Equals(Location::NoLocation())) { 120 scratch = (kind == Location::kRegister) ? Location::RegisterLocation(scratch_index_) 124 return scratch;
|
H A D | code_generator_arm64.cc | 452 Location scratch = GetScratchLocation(kind); local 453 if (!scratch.Equals(Location::NoLocation())) { 454 return scratch; 458 scratch = LocationFrom(vixl_temps_.AcquireX()); 461 scratch = LocationFrom(vixl_temps_.AcquireD()); 463 AddScratchLocation(scratch); 464 return scratch; 1406 // We use a block to end the scratch scope before the write barrier, thus
|
/art/compiler/utils/x86/ |
H A D | assembler_x86.cc | 1837 X86ManagedRegister scratch = mscratch.AsX86(); local 1838 CHECK(scratch.IsCpuRegister()); 1839 leal(scratch.AsCpuRegister(), Address(ESP, fr_offs)); 1840 fs()->movl(Address::Absolute(thr_offs), scratch.AsCpuRegister()); 1848 FrameOffset /*in_off*/, ManagedRegister /*scratch*/) { 1986 X86ManagedRegister scratch = mscratch.AsX86(); local 1987 CHECK(scratch.IsCpuRegister()); 1988 movl(scratch.AsCpuRegister(), Address(ESP, src)); 1989 movl(Address(ESP, dest), scratch.AsCpuRegister()); 1995 X86ManagedRegister scratch local 2004 X86ManagedRegister scratch = mscratch.AsX86(); local 2013 X86ManagedRegister scratch = mscratch.AsX86(); local 2018 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4); local 2030 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument 2040 Register scratch = mscratch.AsX86().AsCpuRegister(); local 2044 movl(Address(ESP, dest), scratch); local 2047 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument 2058 Register scratch = mscratch.AsX86().AsCpuRegister(); local 2096 X86ManagedRegister scratch = mscratch.AsX86(); local 2144 Register scratch = mscratch.AsX86().AsCpuRegister(); local 2160 X86ManagedRegister scratch = mscratch.AsX86(); local [all...] |
H A D | assembler_x86.h | 523 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 525 void StoreImmediateToThread32(ThreadOffset<4> dest, uint32_t imm, ManagedRegister scratch) 529 ManagedRegister scratch) OVERRIDE; 534 ManagedRegister scratch) OVERRIDE; 554 ManagedRegister scratch) OVERRIDE; 556 void CopyRawPtrToThread32(ThreadOffset<4> thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 559 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 561 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 563 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 566 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, [all...] |
/art/compiler/utils/x86_64/ |
H A D | assembler_x86_64.h | 651 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 653 void StoreImmediateToThread64(ThreadOffset<8> dest, uint32_t imm, ManagedRegister scratch) 657 ManagedRegister scratch) OVERRIDE; 662 ManagedRegister scratch) OVERRIDE; 682 ManagedRegister scratch) OVERRIDE; 684 void CopyRawPtrToThread64(ThreadOffset<8> thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 687 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 689 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 691 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 694 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, [all...] |
H A D | assembler_x86_64.cc | 2520 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 2521 CHECK(scratch.IsCpuRegister()); 2522 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), fr_offs)); 2523 gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister()); 2531 FrameOffset /*in_off*/, ManagedRegister /*scratch*/) { 2671 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 2672 CHECK(scratch.IsCpuRegister()); 2673 movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), src)); 2674 movl(Address(CpuRegister(RSP), dest), scratch.AsCpuRegister()); 2680 X86_64ManagedRegister scratch local 2689 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 2697 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 2702 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4); local 2714 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument 2724 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 2731 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument 2742 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 2786 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 2834 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 2848 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local [all...] |