/art/compiler/utils/arm/ |
H A D | jni_macro_assembler_arm_vixl.h | 67 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 71 ManagedRegister scratch) OVERRIDE; 78 ManagedRegister scratch) OVERRIDE; 103 ManagedRegister scratch) OVERRIDE; 107 ManagedRegister scratch) OVERRIDE; 109 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 111 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 116 ManagedRegister scratch, 122 ManagedRegister scratch, 128 ManagedRegister scratch, 207 ArmException(ArmManagedRegister scratch, size_t stack_adjust) argument [all...] |
H A D | jni_macro_assembler_arm.h | 68 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 72 ManagedRegister scratch) OVERRIDE; 77 ManagedRegister scratch) OVERRIDE; 98 ManagedRegister scratch) OVERRIDE; 100 void CopyRawPtrToThread(ThreadOffset32 thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 103 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 105 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 107 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 110 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, 113 void Copy(FrameOffset dest, FrameOffset src_base, Offset src_offset, ManagedRegister scratch, [all...] |
H A D | jni_macro_assembler_arm.cc | 37 ArmExceptionSlowPath(ArmManagedRegister scratch, size_t stack_adjust) argument 38 : scratch_(scratch), stack_adjust_(stack_adjust) { 229 ArmManagedRegister scratch = mscratch.AsArm(); local 231 __ LoadFromOffset(kLoadWord, scratch.AsCoreRegister(), SP, in_off.Int32Value()); 232 __ StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value() + sizeof(uint32_t)); 236 ArmManagedRegister scratch = mscratch.AsArm(); local 237 __ LoadFromOffset(kLoadWord, scratch.AsCoreRegister(), SP, src.Int32Value()); 238 __ StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value()); 280 ArmManagedRegister scratch = mscratch.AsArm(); local 281 CHECK(scratch 326 ArmManagedRegister scratch = mscratch.AsArm(); local 335 ArmManagedRegister scratch = mscratch.AsArm(); local 344 ArmManagedRegister scratch = mscratch.AsArm(); local 404 ArmManagedRegister scratch = mscratch.AsArm(); local 423 Register scratch = mscratch.AsArm().AsCoreRegister(); local 434 Register scratch = mscratch.AsArm().AsCoreRegister(); local 458 Register scratch = mscratch.AsArm().AsCoreRegister(); local 505 ArmManagedRegister scratch = mscratch.AsArm(); local 548 ArmManagedRegister scratch = mscratch.AsArm(); local 560 ArmManagedRegister scratch = mscratch.AsArm(); local 586 ArmManagedRegister scratch = mscratch.AsArm(); local [all...] |
H A D | jni_macro_assembler_arm_vixl.cc | 212 ArmManagedRegister scratch = mscratch.AsArm(); local 215 temps.Exclude(scratch.AsVIXLRegister()); 216 asm_.LoadFromOffset(kLoadWord, scratch.AsVIXLRegister(), sp, in_off.Int32Value()); 217 asm_.StoreToOffset(kStoreWord, scratch.AsVIXLRegister(), sp, dest.Int32Value() + 4); 223 ArmManagedRegister scratch = mscratch.AsArm(); local 225 temps.Exclude(scratch.AsVIXLRegister()); 226 asm_.LoadFromOffset(kLoadWord, scratch.AsVIXLRegister(), sp, src.Int32Value()); 227 asm_.StoreToOffset(kStoreWord, scratch.AsVIXLRegister(), sp, dest.Int32Value()); 261 ManagedRegister scratch) { 262 ArmManagedRegister mscratch = scratch 259 StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) argument 291 ArmManagedRegister scratch = mscratch.AsArm(); local 308 ArmManagedRegister scratch = mscratch.AsArm(); local 372 Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) argument 487 ArmManagedRegister scratch = mscratch.AsArm(); local 534 ArmManagedRegister scratch = mscratch.AsArm(); local 548 ArmManagedRegister scratch = mscratch.AsArm(); local 580 ArmManagedRegister scratch = m_scratch.AsArm(); local [all...] |
/art/compiler/utils/ |
H A D | jni_macro_assembler.h | 79 virtual void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) = 0; 83 ManagedRegister scratch) = 0; 90 ManagedRegister scratch) = 0; 115 ManagedRegister scratch) = 0; 119 ManagedRegister scratch) = 0; 121 virtual void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) = 0; 123 virtual void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) = 0; 128 ManagedRegister scratch, 134 ManagedRegister scratch, 140 ManagedRegister scratch, [all...] |
H A D | swap_space_test.cc | 35 ScratchFile scratch; local 36 int fd = scratch.GetFd(); 37 unlink(scratch.GetFilename().c_str()); 70 scratch.Close();
|
/art/compiler/utils/x86/ |
H A D | jni_macro_assembler_x86.h | 62 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 66 ManagedRegister scratch) OVERRIDE; 71 ManagedRegister scratch) OVERRIDE; 92 ManagedRegister scratch) OVERRIDE; 94 void CopyRawPtrToThread(ThreadOffset32 thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 97 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 99 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 101 void Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister scratch, 104 void Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, 107 void Copy(FrameOffset dest, FrameOffset src_base, Offset src_offset, ManagedRegister scratch, [all...] |
H A D | jni_macro_assembler_x86.cc | 169 X86ManagedRegister scratch = mscratch.AsX86(); local 170 CHECK(scratch.IsCpuRegister()); 171 __ leal(scratch.AsCpuRegister(), Address(ESP, fr_offs)); 172 __ fs()->movl(Address::Absolute(thr_offs), scratch.AsCpuRegister()); 182 ManagedRegister /*scratch*/) { 323 X86ManagedRegister scratch = mscratch.AsX86(); local 324 CHECK(scratch.IsCpuRegister()); 325 __ movl(scratch.AsCpuRegister(), Address(ESP, src)); 326 __ movl(Address(ESP, dest), scratch.AsCpuRegister()); 332 X86ManagedRegister scratch local 341 X86ManagedRegister scratch = mscratch.AsX86(); local 350 X86ManagedRegister scratch = mscratch.AsX86(); local 355 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4); local 370 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument 386 Register scratch = mscratch.AsX86().AsCpuRegister(); local 390 __ movl(Address(ESP, dest), scratch); local 393 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument 411 Register scratch = mscratch.AsX86().AsCpuRegister(); local 450 X86ManagedRegister scratch = mscratch.AsX86(); local 498 Register scratch = mscratch.AsX86().AsCpuRegister(); local 514 X86ManagedRegister scratch = mscratch.AsX86(); local [all...] |
/art/compiler/utils/x86_64/ |
H A D | jni_macro_assembler_x86_64.h | 63 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 67 ManagedRegister scratch) OVERRIDE; 74 ManagedRegister scratch) OVERRIDE; 97 ManagedRegister scratch) OVERRIDE; 99 void CopyRawPtrToThread(ThreadOffset64 thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 102 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 104 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 109 ManagedRegister scratch, 115 ManagedRegister scratch, 121 ManagedRegister scratch, [all...] |
H A D | jni_macro_assembler_x86_64.cc | 208 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 209 CHECK(scratch.IsCpuRegister()); 210 __ leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), fr_offs)); 211 __ gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister()); 221 ManagedRegister /*scratch*/) { 373 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 374 CHECK(scratch.IsCpuRegister()); 375 __ movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), src)); 376 __ movl(Address(CpuRegister(RSP), dest), scratch.AsCpuRegister()); 382 X86_64ManagedRegister scratch local 391 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 401 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 406 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4); local 421 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument 437 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 444 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument 462 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 507 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local 555 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister(); local 570 X86_64ManagedRegister scratch = mscratch.AsX86_64(); local [all...] |
/art/compiler/utils/arm64/ |
H A D | jni_macro_assembler_arm64.cc | 55 void Arm64JNIMacroAssembler::GetCurrentThread(FrameOffset offset, ManagedRegister /* scratch */) { 161 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 162 CHECK(scratch.IsXRegister()) << scratch; 163 LoadImmediate(scratch.AsXRegister(), imm); 164 StoreWToOffset(kStoreWord, scratch.AsOverlappingWRegister(), SP, 171 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 172 CHECK(scratch.IsXRegister()) << scratch; 173 AddConstant(scratch 189 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 364 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 373 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 380 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 392 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 411 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 433 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 464 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 537 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 545 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 590 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 625 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local [all...] |
H A D | jni_macro_assembler_arm64.h | 69 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) OVERRIDE; 72 ManagedRegister scratch) OVERRIDE; 77 ManagedRegister scratch) OVERRIDE; 94 ManagedRegister scratch) OVERRIDE; 95 void CopyRawPtrToThread(ThreadOffset64 thr_offs, FrameOffset fr_offs, ManagedRegister scratch) 97 void CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister scratch) OVERRIDE; 98 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 102 ManagedRegister scratch, 107 ManagedRegister scratch, 112 ManagedRegister scratch, 183 Arm64Exception(Arm64ManagedRegister scratch, size_t stack_adjust) argument [all...] |
H A D | assembler_arm64.cc | 64 Arm64ManagedRegister scratch = m_scratch.AsArm64(); local 66 CHECK(scratch.IsXRegister()) << scratch; 67 // Remove base and scratch form the temp list - higher level API uses IP1, IP0. 69 temps.Exclude(reg_x(base.AsXRegister()), reg_x(scratch.AsXRegister())); 70 ___ Ldr(reg_x(scratch.AsXRegister()), MEM_OP(reg_x(base.AsXRegister()), offs.Int32Value())); 71 ___ Br(reg_x(scratch.AsXRegister()));
|
/art/compiler/optimizing/ |
H A D | parallel_move_resolver.cc | 72 // Eliminate the move, in case following moves need a scratch register. 264 int scratch = -1; local 267 scratch = reg; 272 if (scratch == -1) { 274 scratch = if_scratch; 279 return scratch; 381 for (Location scratch : scratches_) { 382 CHECK(!loc.Equals(scratch)); 402 // in the move graph. We use scratch location to resolve cycles, also 414 // resolve as (1 -> scratch) ( 458 Location scratch = AllocateScratchLocationFor(kind); local 495 Location scratch = scratches_[i]; local [all...] |
H A D | parallel_move_test.cc | 115 Location scratch = GetScratchLocation(kind); variable 116 if (scratch.Equals(Location::NoLocation())) { 120 scratch = (kind == Location::kRegister) ? Location::RegisterLocation(scratch_index_) 124 return scratch;
|
H A D | code_generator_arm64.cc | 1497 Location scratch = GetScratchLocation(kind); local 1498 if (!scratch.Equals(Location::NoLocation())) { 1499 return scratch; 1503 scratch = LocationFrom(vixl_temps_.AcquireX()); 1506 scratch = LocationFrom(codegen_->GetGraph()->HasSIMD() 1510 AddScratchLocation(scratch); 1511 return scratch; 1733 // Allocate a scratch register from the VIXL pool, querying first into 1872 // Use any scratch register (a core or a floating-point one) 1873 // from VIXL scratch registe [all...] |
/art/compiler/utils/mips64/ |
H A D | assembler_mips64.cc | 2893 Mips64ManagedRegister scratch = mscratch.AsMips64(); 2894 CHECK(scratch.IsGpuRegister()) << scratch; 2895 LoadConst32(scratch.AsGpuRegister(), imm); 2896 StoreToOffset(kStoreWord, scratch.AsGpuRegister(), SP, dest.Int32Value()); 2902 Mips64ManagedRegister scratch = mscratch.AsMips64(); 2903 CHECK(scratch.IsGpuRegister()) << scratch; 2904 Daddiu64(scratch.AsGpuRegister(), SP, fr_offs.Int32Value()); 2905 StoreToOffset(kStoreDoubleword, scratch [all...] |
H A D | assembler_mips64.h | 395 explicit Mips64ExceptionSlowPath(Mips64ManagedRegister scratch, size_t stack_adjust) argument 396 : scratch_(scratch), stack_adjust_(stack_adjust) {}
|
/art/compiler/utils/mips/ |
H A D | assembler_mips.cc | 2833 // as a scratch register (see InterproceduralScratchRegister()). 3448 MipsManagedRegister scratch = mscratch.AsMips(); local 3449 CHECK(scratch.IsCoreRegister()) << scratch; 3450 LoadConst32(scratch.AsCoreRegister(), imm); 3451 StoreToOffset(kStoreWord, scratch.AsCoreRegister(), SP, dest.Int32Value()); 3457 MipsManagedRegister scratch = mscratch.AsMips(); local 3458 CHECK(scratch.IsCoreRegister()) << scratch; 3459 Addiu32(scratch 3471 MipsManagedRegister scratch = mscratch.AsMips(); local 3557 MipsManagedRegister scratch = mscratch.AsMips(); local 3566 MipsManagedRegister scratch = mscratch.AsMips(); local 3577 MipsManagedRegister scratch = mscratch.AsMips(); local 3586 MipsManagedRegister scratch = mscratch.AsMips(); local 3602 Register scratch = mscratch.AsMips().AsCoreRegister(); local 3610 Register scratch = mscratch.AsMips().AsCoreRegister(); local 3628 Register scratch = mscratch.AsMips().AsCoreRegister(); local 3680 MipsManagedRegister scratch = mscratch.AsMips(); local 3726 MipsManagedRegister scratch = mscratch.AsMips(); local 3737 MipsManagedRegister scratch = mscratch.AsMips(); local 3763 MipsManagedRegister scratch = mscratch.AsMips(); local [all...] |
H A D | assembler_mips.h | 162 explicit MipsExceptionSlowPath(MipsManagedRegister scratch, size_t stack_adjust) argument 163 : scratch_(scratch), stack_adjust_(stack_adjust) {}
|