/art/compiler/dex/quick/arm64/ |
H A D | int_arm64.cc | 1453 uint32_t fp_reg_mask, int frame_size) { 1454 m2l->OpRegRegImm(kOpSub, rs_sp, rs_sp, frame_size); 1461 int spill_offset = frame_size - (core_count + fp_count) * kArm64PointerSize; 1467 int spill_offset = frame_size - (core_count * kArm64PointerSize); 1471 return frame_size; 1475 uint32_t fp_reg_mask, int frame_size) { 1590 int frame_size) { 1597 if ((RoundUp(frame_size, 8) / 8 <= 63)) { 1598 return SpillRegsPreSub(this, base, core_reg_mask, fp_reg_mask, frame_size); 1600 return SpillRegsPreIndexed(this, base, core_reg_mask, fp_reg_mask, frame_size); 1452 SpillRegsPreSub(Arm64Mir2Lir* m2l, RegStorage base, uint32_t core_reg_mask, uint32_t fp_reg_mask, int frame_size) argument 1474 SpillRegsPreIndexed(Arm64Mir2Lir* m2l, RegStorage base, uint32_t core_reg_mask, uint32_t fp_reg_mask, int frame_size) argument 1589 SpillRegs(RegStorage base, uint32_t core_reg_mask, uint32_t fp_reg_mask, int frame_size) argument 1636 UnspillRegs(RegStorage base, uint32_t core_reg_mask, uint32_t fp_reg_mask, int frame_size) argument [all...] |
/art/compiler/jni/quick/ |
H A D | jni_compiler.cc | 47 size_t frame_size, size_t out_arg_size); 101 const size_t frame_size(main_jni_conv->FrameSize()); 103 __ BuildFrame(frame_size, mr_conv->MethodRegister(), callee_save_regs, mr_conv->EntrySpills()); 106 mr_conv->ResetIterator(FrameOffset(frame_size)); 134 CHECK_LT(handle_scope_offset.Uint32Value(), frame_size); 151 CHECK_LT(handle_scope_offset.Uint32Value(), frame_size); 246 mr_conv->ResetIterator(FrameOffset(frame_size + main_out_arg_size)); 257 mr_conv->ResetIterator(FrameOffset(frame_size + main_out_arg_size)); 268 CopyParameter(jni_asm.get(), mr_conv.get(), main_jni_conv.get(), frame_size, main_out_arg_size); 272 mr_conv->ResetIterator(FrameOffset(frame_size 446 CopyParameter(Assembler* jni_asm, ManagedRuntimeCallingConvention* mr_conv, JniCallingConvention* jni_conv, size_t frame_size, size_t out_arg_size) argument [all...] |
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.cc | 639 void Arm64Assembler::BuildFrame(size_t frame_size, ManagedRegister method_reg, argument 642 CHECK_ALIGNED(frame_size, kStackAlignment); 650 CHECK_GT(frame_size, kJniRefSpillRegsSize * kFramePointerSize); 651 IncreaseFrameSize(frame_size); 657 size_t reg_offset = frame_size; 690 int32_t offset = frame_size + sizeof(StackReference<mirror::ArtMethod>); 713 void Arm64Assembler::RemoveFrame(size_t frame_size, const std::vector<ManagedRegister>& callee_save_regs) { argument 714 CHECK_ALIGNED(frame_size, kStackAlignment); 718 CHECK_GT(frame_size, kJniRefSpillRegsSize * kFramePointerSize); 728 size_t reg_offset = frame_size; [all...] |
/art/compiler/utils/mips/ |
H A D | assembler_mips.cc | 541 void MipsAssembler::BuildFrame(size_t frame_size, ManagedRegister method_reg, argument 544 CHECK_ALIGNED(frame_size, kStackAlignment); 547 IncreaseFrameSize(frame_size); 550 int stack_offset = frame_size - kFramePointerSize; 564 StoreToOffset(kStoreWord, reg, SP, frame_size + kFramePointerSize + (i * kFramePointerSize)); 568 void MipsAssembler::RemoveFrame(size_t frame_size, argument 570 CHECK_ALIGNED(frame_size, kStackAlignment); 573 int stack_offset = frame_size - (callee_save_regs.size() * kFramePointerSize) - kFramePointerSize; 582 DecreaseFrameSize(frame_size);
|
/art/compiler/utils/x86/ |
H A D | assembler_x86.cc | 1412 void X86Assembler::BuildFrame(size_t frame_size, ManagedRegister method_reg, argument 1415 CHECK_ALIGNED(frame_size, kStackAlignment); 1420 addl(ESP, Immediate(-frame_size + (spill_regs.size() * kFramePointerSize) + 1425 movl(Address(ESP, frame_size + sizeof(StackReference<mirror::ArtMethod>) + 1431 void X86Assembler::RemoveFrame(size_t frame_size, argument 1433 CHECK_ALIGNED(frame_size, kStackAlignment); 1434 addl(ESP, Immediate(frame_size - (spill_regs.size() * kFramePointerSize) -
|
/art/compiler/utils/x86_64/ |
H A D | assembler_x86_64.cc | 1719 void X86_64Assembler::BuildFrame(size_t frame_size, ManagedRegister method_reg, argument 1722 CHECK_ALIGNED(frame_size, kStackAlignment); 1732 int64_t rest_of_frame = static_cast<int64_t>(frame_size) 1754 movq(Address(CpuRegister(RSP), frame_size + spill.getSpillOffset()), 1758 movl(Address(CpuRegister(RSP), frame_size + spill.getSpillOffset()), spill.AsX86_64().AsCpuRegister()); 1762 movsd(Address(CpuRegister(RSP), frame_size + spill.getSpillOffset()), spill.AsX86_64().AsXmmRegister()); 1765 movss(Address(CpuRegister(RSP), frame_size + spill.getSpillOffset()), spill.AsX86_64().AsXmmRegister()); 1771 void X86_64Assembler::RemoveFrame(size_t frame_size, argument 1773 CHECK_ALIGNED(frame_size, kStackAlignment); 1776 int64_t offset = static_cast<int64_t>(frame_size) [all...] |
H A D | assembler_x86_64_test.cc | 256 size_t frame_size = 10 * kStackAlignment; local 265 ssize_t displacement = static_cast<ssize_t>(frame_size) - (spill_regs.size() * 8 + 8); 270 str << "movq %rax, " << frame_size + 0 << "(%rsp)\n"; local 271 str << "movq %rbx, " << frame_size + 8 << "(%rsp)\n"; local 272 str << "movsd %xmm1, " << frame_size + 16 << "(%rsp)\n"; local 289 size_t frame_size = 10 * kStackAlignment; local 295 ssize_t displacement = static_cast<ssize_t>(frame_size) - spill_regs.size() * 8 - 8;
|
/art/runtime/ |
H A D | elf_file.cc | 1495 static DebugInfoIterator* Create(DebugInfoHeader* header, size_t frame_size, argument 1497 std::unique_ptr<DebugInfoIterator> iter(new DebugInfoIterator(header, frame_size, abbrev)); 1544 DebugInfoIterator(DebugInfoHeader* header, size_t frame_size, DebugAbbrev* abbrev) argument 1546 last_entry_(reinterpret_cast<byte*>(header) + frame_size),
|
H A D | stack.cc | 506 size_t frame_size = method->GetFrameSizeInBytes(); local 507 CHECK_NE(frame_size, 0u); 516 CHECK_LE(frame_size, kMaxExpectedFrameSize); 518 CHECK_LT(return_pc_offset, frame_size); 549 size_t frame_size = method->GetFrameSizeInBytes(); local 551 size_t return_pc_offset = method->GetReturnPcOffsetInBytes(frame_size); 582 byte* next_frame = reinterpret_cast<byte*>(cur_quick_frame_) + frame_size;
|
H A D | stack.h | 541 uintptr_t* CalleeSaveAddress(int num, size_t frame_size) const 546 reinterpret_cast<byte*>(cur_quick_frame_) + frame_size - ((num + 1) * kPointerSize); 614 uint32_t core_spills, uint32_t fp_spills, size_t frame_size, 616 int offset = GetVRegOffset(code_item, core_spills, fp_spills, frame_size, vreg, kRuntimeISA); 673 size_t frame_size, int reg, InstructionSet isa) { 674 DCHECK_EQ(frame_size & (kStackAlignment - 1), 0U); 681 int locals_start = frame_size - spill_size - num_regs * sizeof(uint32_t); 701 return frame_size + ((reg - num_regs) * sizeof(uint32_t)) + 612 GetVRegAddr(StackReference<mirror::ArtMethod>* cur_quick_frame, const DexFile::CodeItem* code_item, uint32_t core_spills, uint32_t fp_spills, size_t frame_size, uint16_t vreg) const argument 671 GetVRegOffset(const DexFile::CodeItem* code_item, uint32_t core_spills, uint32_t fp_spills, size_t frame_size, int reg, InstructionSet isa) argument
|
/art/runtime/entrypoints/quick/ |
H A D | quick_field_entrypoints.cc | 251 constexpr size_t frame_size = GetCalleeSaveFrameSize(kRuntimeISA, Runtime::kRefsOnly); local 254 reinterpret_cast<uint8_t*>(sp) + frame_size)->AsMirrorPtr();
|
/art/runtime/mirror/ |
H A D | art_method-inl.h | 394 size_t frame_size = RoundUp(callee_info.FrameSizeInBytes() + scope_size local 399 return QuickMethodFrameInfo(frame_size, callee_info.CoreSpillMask(), callee_info.FpSpillMask());
|