/art/runtime/base/ |
H A D | debug_stack.h | 113 : ref_(ref) { 117 : ref_(other.ref_) { 121 CHECK(ref_ == other.ref_); 129 ref_->CheckTop(); 133 DebugStackReferenceImpl<kIsDebug>* ref_; member in class:art::DebugStackIndirectTopRefImpl
|
/art/compiler/optimizing/ |
H A D | code_generator_mips.cc | 530 : SlowPathCodeMIPS(instruction), ref_(ref), entrypoint_(entrypoint) { 538 Register ref_reg = ref_.AsRegister<Register>(); 595 const Location ref_; member in class:art::mips::ReadBarrierMarkSlowPathMIPS 621 ref_(ref), 634 Register ref_reg = ref_.AsRegister<Register>(); 759 const Location ref_; member in class:art::mips::ReadBarrierMarkAndUpdateFieldSlowPathMIPS 781 ref_(ref), 891 parallel_move.AddMove(ref_, 925 size_t ref = static_cast<int>(ref_.AsRegister<Register>()); 945 const Location ref_; member in class:art::mips::ReadBarrierForHeapReferenceSlowPathMIPS [all...] |
H A D | code_generator_mips64.cc | 476 : SlowPathCodeMIPS64(instruction), ref_(ref), entrypoint_(entrypoint) { 484 GpuRegister ref_reg = ref_.AsRegister<GpuRegister>(); 540 const Location ref_; member in class:art::mips64::ReadBarrierMarkSlowPathMIPS64 566 ref_(ref), 579 GpuRegister ref_reg = ref_.AsRegister<GpuRegister>(); 694 const Location ref_; member in class:art::mips64::ReadBarrierMarkAndUpdateFieldSlowPathMIPS64 716 ref_(ref), 824 parallel_move.AddMove(ref_, 860 size_t ref = static_cast<int>(ref_.AsRegister<GpuRegister>()); 880 const Location ref_; member in class:art::mips64::ReadBarrierForHeapReferenceSlowPathMIPS64 [all...] |
H A D | code_generator_arm64.cc | 676 : SlowPathCodeARM64(instruction), ref_(ref), entrypoint_(entrypoint) { 689 DCHECK_NE(ref_.reg(), LR); 690 DCHECK_NE(ref_.reg(), WSP); 691 DCHECK_NE(ref_.reg(), WZR); 694 DCHECK_NE(ref_.reg(), IP0); 695 DCHECK(0 <= ref_.reg() && ref_.reg() < kNumberOfWRegisters) << ref_.reg(); 716 CodeGenerator::GetReadBarrierMarkEntryPointsOffset<kArm64PointerSize>(ref_.reg()); 723 const Location ref_; member in class:art::arm64::ReadBarrierMarkSlowPathBaseARM64 1321 const Location ref_; member in class:art::arm64::ReadBarrierForHeapReferenceSlowPathARM64 [all...] |
H A D | code_generator_arm.cc | 652 : SlowPathCodeARM(instruction), ref_(ref), entrypoint_(entrypoint) { 661 Register ref_reg = ref_.AsRegister<Register>(); 701 const Location ref_; member in class:art::arm::ReadBarrierMarkSlowPathBaseARM 739 Register ref_reg = ref_.AsRegister<Register>(); 800 Register ref_reg = ref_.AsRegister<Register>(); 862 instruction_, ref_, obj_, offset_, index_, scale_factor_, /* needs_null_check */ false); 943 Register ref_reg = ref_.AsRegister<Register>(); 983 instruction_, ref_, obj_, offset_, index_, scale_factor_, /* needs_null_check */ false); 1113 ref_(ref), 1228 parallel_move.AddMove(ref_, 1276 const Location ref_; member in class:art::arm::ReadBarrierForHeapReferenceSlowPathARM [all...] |
H A D | code_generator_arm_vixl.cc | 676 : SlowPathCodeARMVIXL(instruction), ref_(ref), entrypoint_(entrypoint) { 685 vixl32::Register ref_reg = RegisterFrom(ref_); 725 const Location ref_; member in class:art::arm::ReadBarrierMarkSlowPathBaseARMVIXL 762 DCHECK(ref_.IsRegister()) << ref_; 763 DCHECK(!locations->GetLiveRegisters()->ContainsCoreRegister(ref_.reg())) << ref_.reg(); 822 vixl32::Register ref_reg = RegisterFrom(ref_); 890 instruction_, ref_, obj_, offset_, index_, scale_factor_, /* needs_null_check */ false); 972 vixl32::Register ref_reg = RegisterFrom(ref_); 1320 const Location ref_; member in class:art::arm::ReadBarrierForHeapReferenceSlowPathARMVIXL [all...] |
H A D | code_generator_x86.cc | 461 ref_(ref), 470 Register ref_reg = ref_.AsRegister<Register>(); 520 const Location ref_; member in class:art::x86::ReadBarrierMarkSlowPathX86 521 // Should the reference in `ref_` be unpoisoned prior to marking it? 546 ref_(ref), 558 Register ref_reg = ref_.AsRegister<Register>(); 678 const Location ref_; member in class:art::x86::ReadBarrierMarkAndUpdateFieldSlowPathX86 684 // Should the reference in `ref_` be unpoisoned prior to marking it? 703 ref_(ref), 813 parallel_move.AddMove(ref_, 861 const Location ref_; member in class:art::x86::ReadBarrierForHeapReferenceSlowPathX86 [all...] |
H A D | code_generator_x86_64.cc | 474 ref_(ref), 483 CpuRegister ref_cpu_reg = ref_.AsRegister<CpuRegister>(); 534 const Location ref_; member in class:art::x86_64::ReadBarrierMarkSlowPathX86_64 535 // Should the reference in `ref_` be unpoisoned prior to marking it? 561 ref_(ref), 576 CpuRegister ref_cpu_reg = ref_.AsRegister<CpuRegister>(); 698 const Location ref_; member in class:art::x86_64::ReadBarrierMarkAndUpdateFieldSlowPathX86_64 704 // Should the reference in `ref_` be unpoisoned prior to marking it? 724 ref_(ref), 831 parallel_move.AddMove(ref_, 884 const Location ref_; member in class:art::x86_64::ReadBarrierForHeapReferenceSlowPathX86_64 [all...] |